Hamming distance completeness and sparse matrix multiplication

نویسندگان

  • Daniel Graf
  • Karim Labib
  • Przemyslaw Uznanski
چکیده

We investigate relations between (+, ) vector products for binary integer functions . We show that there exists a broad class of products that are equivalent under one-to-polylog reductions to the computation of the Hamming distance. Examples include, but are not exhausted by: the dominance product, the threshold product and `2p+1 distances for constant p. Our result has the following consequences: • The following problems are of the same complexity (up to polylog factors) for n given vectors in Z: computing AllPairsHammingDistances, AllPairsL2p+1Distances for constant p, AllPairsThresholdProducts and AllPairsDominanceProducts. As a consequence, Yuster’s (SODA’09) algorithm improves not only Matoušek’s (IPL’91), but also the results of Indyk, Lewenstein, Lipsky and Porat (ICALP’04) and Min, Kao and Zhu (COCOON’09). Moreover, we obtain by reduction algorithms for All Pairs `3, `5, . . . Distances with the same runtime. • The following problems are of the same complexity (up to polylog factors) for a given text of length n and a pattern of length m: HammingDistancePatternMatching, LessThanPatternMatching, ThresholdPatternMatching and L2p+1PatternMatching for constant p. Thus it is no coincidence that for all those problems the current best upper bounds are Õ(nm) time due to results of Abrahamson (SICOMP’87), Amir and Farach (Ann. Math. Artif. Intell.’91), Atallah and Duket (IPL’11), Clifford, Clifford and Iliopoulous (CPM’05) and Amir, Lipsky, Porat and Umanski (CPM’05). The algorithms for `3, `5, . . . Pattern Matchings we obtained by our reduction are new. Additionally, we show that the complexity of AllPairsHammingDistances (and thus of other aforementioned AllPairsproblems) is within a polylog factor from Sparse(n, d, n;nd, nd), where Sparse(a, b, c;m1,m2) is the time of multiplying sparse matrices of size a × b and b × c, with m1 and m2 nonzero entries, respectively. This means that the current upperbounds by Yuster cannot be improved without improving the sparse matrix multiplication algorithm by Yuster and Zwick (ACM TALG’05) and vice versa. ar X iv :1 71 1. 03 88 7v 1 [ cs .D S] 1 0 N ov 2 01 7

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Fault-Tolerant Dot-Product Engines

Coding schemes are presented that provide the ability to correct and detect computational errors while using dotproduct engines for integer vector–matrix multiplication. Both the L1-metric and the Hamming metric are considered.

متن کامل

A Clustering-Based Matrix Multiplication Algorithm

We present a simple matrix multiplication algorithm that multiplies two input matrices with rows (in one matrix) and columns (in the other matrix) within a small diameter d (distances are measured using the Hamming distance). This algorithm runs in time O(dn) for matrices of size n × n. We then propose a more general clusteringbased matrix multiplication algorithm. For a given integer k ≤ n, th...

متن کامل

Upper and Lower Bounds on the Cost of a Map-Reduce Computation

In this paper we study the tradeoff between parallelism and communication cost in a map-reduce computation. For any problem that is not “embarrassingly parallel,” the finer we partition the work of the reducers so that more parallelism can be extracted, the greater will be the total communication between mappers and reducers. We introduce a model of problems that can be solved in a single round...

متن کامل

An Advanced Method for Joint Scalar Multiplications on Memory Constraint Devices

One of the most frequent operations in modern cryptosystems is a multi-scalar multiplication with two scalars. Common methods to compute it are the Shamir method and the Interleave method whereas their speed mainly depends on the (joint) Hamming weight of the scalars. To increase the speed, the scalars are usually deployed using some general representation which provides a lower (joint) Hamming...

متن کامل

Upper and Lower Bounds for Dynamic Data Structures on Strings

We consider a range of simply stated dynamic data structure problems on strings. An update changes one symbol in the input and a query asks us to compute some function of the pattern of length m and a substring of a longer text. We give both conditional and unconditional lower bounds for variants of exact matching with wildcards, inner product, and Hamming distance computation via a sequence of...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • CoRR

دوره abs/1711.03887  شماره 

صفحات  -

تاریخ انتشار 2017